Published on in Vol 5 (2026)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/54956, first published .
Enhancing COVID-19 Screening Models With Epidemiological and Mobility Features: Machine-Learning Model Study

Enhancing COVID-19 Screening Models With Epidemiological and Mobility Features: Machine-Learning Model Study

Enhancing COVID-19 Screening Models With Epidemiological and Mobility Features: Machine-Learning Model Study

1Department of Digital Health, Samsung Advanced Institute for Health Sciences and Technology (SAIHST), Sungkyunkwan University, Seoul, Republic of Korea

2Mobile Doctor Co., Ltd, Seoul, Republic of Korea

3Department of Biomedical Software Engineering, The Catholic University of Korea, 43 Jibong-ro, Wonmi-gu, Bucheon-si, Gyeonggi-do, Republic of Korea

*these authors contributed equally

Corresponding Author:

Hyo Jung Kim, PhD


Background: Despite the significant post–COVID-19 pandemic surge in research using symptom data and machine learning (ML) for patient screening, data on patient trajectories and epidemiological conditions, although crucial, have remained underused.

Objective: This study aimed to enhance the performance of ML models for COVID-19 screening by incorporating mobility and epidemic information in addition to patient symptom data.

Methods: Data, including daily self-reported symptoms, location information, and test results, were collected from 48,798 individuals using a smartphone app. These data were then combined with Our World in Data and national government epidemic information to train 5 ML-based screening models to classify patient infection status. The models were logistic regression, extreme gradient boosting, light gradient boosting machine, tabular data network, and Google AutoML.

Results: The addition of mobility and epidemic data significantly improved the performance of all 5 models. The highest area under the receiver operating characteristic curve score increased from 0.8712 without mobility and epidemic data to 0.9104 with mobility and epidemic data. This highlights the considerable impact of external information on enhancing the performance of ML models.

Conclusions: This study demonstrated the potential of using mobility and epidemic data, such as location information and epidemic data, in combination with patient symptom data to improve the accuracy of ML models for diagnosing COVID-19. Considering additional contextual information can enhance the ability to screen for COVID-19.

JMIR AI 2026;5:e54956

doi:10.2196/54956

Keywords



Since the onset of the COVID-19 pandemic, rapid and accurate diagnostic methods have become crucial for controlling the spread of the disease. Although polymerase chain reaction (PCR) testing remains the gold standard for diagnosis [1], its limitations with respect to the need for specialized equipment, facilities, and time have hindered its scalability during the early stages of the pandemic [2,3].

In COVID-19’s early stages, screening strategies mainly focused on travel history and contact with infected individuals [4], primarily due to limited PCR capacity, as exemplified by the Centers for Disease Control and Prevention initially targeting international travelers for monitoring [5]. However, as community transmission increased, these methods were insufficient, leading to a shift toward symptom-based testing strategies [6]. This transition is supported by the fact that symptoms serve as primary indicators prompting testing for a range of infectious diseases, including influenza [7,8], Zika [9,10], malaria [11], and Ebola [12]. Particularly in areas lacking laboratory support, symptom-based screening has been recommended as an effective approach, as seen in the case of malaria [11].

Some research has shown promise in identifying potential COVID-19 cases using self-reported data and machine learning (ML) following a symptom-based screening method. Zoabi et al [13] developed an ML model that predicted COVID-19 test results with high accuracy using only 8 binary features, including demographic information and the presence of initial clinical symptoms [12]. Similarly, the COVID symptom study app in Sweden, which analyzed daily symptom reports from participants, developed a symptom-based model that outperformed traditional case notification-based models in predicting hospital admissions during the first wave of the pandemic [14].

However, the overlap of COVID-19 symptoms with those of other respiratory illnesses, such as influenza, has posed significant challenges in accurately identifying cases based on symptoms alone [15-19]. In addition, the presence of asymptomatic individuals infected with COVID-19 further complicates the accurate identification process, making reliance solely on symptom-based methods more challenging [20].

On the other hand, research on mathematical models that predict infectious disease prevalence and understand transmission dynamics, incorporating surveillance data, offers vital insights into public health strategies while also emphasizing the significance of epidemiological factors [21-23]. These models often consider factors such as traveling population, mobility, and contact patterns to better understand disease spread and inform control measures. Choo et al [24] show that combining epidemiological data with individual health data can enhance screening model accuracy. Loo et al [25] show that using mobility data enables the identification and characterization of superspreaders. These findings suggest that integrating community-level data and transmission dynamics into individual screening strategies could potentially enhance the effectiveness of COVID-19 screening as well, bridging the gap between broad epidemiological insights and actionable, individual-level interventions.

Given the constraints of symptom-based screening for COVID-19 and the potential advantages of incorporating epidemiological factors and mobility data, a combined approach could enhance screening efficacy. Previous studies have demonstrated the effectiveness of mobile apps in gathering symptom-based survey data and diagnostic test results [12,13]. Building upon this, we developed the SHINE (Study of Health Information for Next Epidemic; AI/DX Convergence Business Group) app to focus on collecting data for testing our combined approach. The app gathers symptom data via a smartphone interface, with users reporting specific symptoms outlined in the methods section. In addition, the app’s design facilitates the integration of mobility data and local surveillance indices, enabling the creation of a more comprehensive screening tool that considers the nonspecific nature of COVID-19 symptoms and the impact of epidemiological factors on disease spread. The app interacts with patients who have undergone diagnostic tests by enabling them to submit screenshots or photos of their results, which are then validated by the app’s reviewers to ensure data accuracy. This validation process aims to address the limitations of existing patient-generated health data platforms, where diagnosis data are often self-reported and may lack robustness for training ML models. By integrating verified diagnostic test results, the SHINE app aims to furnish a more dependable dataset for training precise COVID-19 screening models.

We postulated that merging symptom-based profiling of COVID-19 with mobility data and local surveillance indices could yield a more precise and resilient screening tool, acknowledging the nonspecificity of COVID-19 symptoms. Thus, our study endeavors to assess the efficacy of this integrated approach in enhancing COVID-19 screening strategies.


Study Design

The main goal of this study is to demonstrate that deep learning models can better classify COVID-19 infections by incorporating additional data sources such as GPS-based mobility data and publicly available COVID-19 metrics alongside daily self-check symptom data, which serves as the primary input. In Figure 1, the overall study design is depicted. We used the SHINE dataset to train and evaluate 5 different ML models for COVID-19 screening. This study adheres to the CONSORT-AI (Consolidated Standards of Reporting Trials–Artificial Intelligence) reporting guideline.

Figure 1. Overview of the study design. AutoML: automated machine learning; LGBM: light gradient boosting machine; SHINE: Study of Health Information for Next Epidemic; TabNet: tabular data network; XGBoost: extreme gradient boosting.

Prior to our main analysis, we conducted a preliminary experiment using a public dataset from Israel [26], which demonstrated potential improvements in the predictive performance of ML models by incorporating readily available COVID-19 epidemiological data (Multimedia Appendix 1). This preliminary work supported our hypothesis that contextual information beyond patient symptoms could enhance screening accuracy.

For our main experiment, we leveraged the SHINE dataset, which contained rich contextual information, including mobility data obtained through GPS tracking. In this experiment, symptom data and demographic information were used as primary variables, and the analysis was extended by incorporating 3 levels of epidemiological data (global, national, and regional) along with mobility data as supplementary contextual variables. The main experiment evaluated the impact of these variables by adding their combinations to the base variables, allowing us to observe how additional contextual information affected the models’ predictive performance.

Data Source

The SHINE dataset was collected through our proprietary platform called SHINE. As part of the platform, the SHINE app enabled daily self-check submissions from users, providing crucial demographic information such as age, sex, underlying health conditions, and mobility patterns obtained through smartphone location data [27]. Over approximately 3 years, we collaboratively developed and maintained the SHINE smartphone app with hospitals and private organizations to gather personal health data from users. Concurrently, an additional database was established by incorporating case statistics from various cities and districts provided by the central and local governments.

For contextual epidemiological information, we incorporated relevant data from Our World in Data, including global and national surveillance data [28].

SHINE Dataset

Self-check symptoms and authorized COVID-19 test results, along with basic information such as biological sex and age, were collected from October 2020 to March 2023 using the SHINE mobile app in South Korea. Some individuals submitted self-reported data without reporting their authorized COVID-19 test results, and vice versa. Notably, as the government’s standards for confirming positive COVID-19 cases varied over time, there was a mix of outcomes from different types of tests, such as PCR and rapid antigen tests. To enhance the reliability of the data, authorized government-issued documents were manually reviewed.

Data from users who previously submitted the results of authorized COVID-19 tests at least once were selected, regardless of whether the test results were positive or negative. Subsequently, the self-checked symptom data of the users were grouped into episodes based on their unique user identifiers. Each episode was then limited to a range of 2 weeks before and after the individual’s authorized COVID-19 test date.

During this process, we identified a subset of users who reported more than 2 authorized COVID-19 test results, along with their corresponding self-checked symptoms. Instead of excluding these users to prevent confounding owing to repeated testing at short intervals, we ensured that the episodes generated by the same individual were kept in the same group when splitting the data into training and testing sets.

Data Extraction and Episode Definition

We defined an “episode” as a 14-day window around each authorized COVID-19 test date. If a user had multiple tests on different dates (eg, 2 weeks apart), each test was treated as a separate episode. For each episode, we extracted the day of self-reported data that was closest to the test date. This meant that each episode ultimately yielded 1 single-day observation representing symptoms, mobility, and epidemiological data.

Some users underwent multiple COVID-19 tests, generating distinct episodes (single-day observations) per test. To prevent data leakage arising from individual-specific patterns (eg, demographics and behaviors) across episodes, all episodes from the same user were exclusively allocated to either the training or test set. This ensured model evaluation reflected generalization to unseen individuals rather than memorization of recurring user traits.

Ethical Considerations

This study was approved by the Institutional Review Board of Sungkyunkwan University (IRB No: SKKU-2023-04-047 and SKKU 2022-11-014). Informed consent was waived because the entire dataset is anonymized, preventing researchers from accessing sensitive personal information and minimizing potential risks. No compensation was provided to participants since only deidentified secondary data were used.


Mobility Data

The mobility data used in this study were obtained from GPS tracking data collected using the SHINE app. The raw GPS positions consisted of latitude and longitude coordinates along with timestamps. Our hypothesis was that increased movement might correlate with a higher risk of COVID-19 infection, so we extracted 3 different metrics to capture distinct aspects of patient mobility:

  1. Number of GPS signals captured daily: this metric reflects the frequency of GPS signal capture, which correlates with the amount of time a patient is traveling.
  2. Daily travel distance: calculated using the Haversine formula, this metric measures the distance traveled each day between recorded GPS points. The Haversine formula is appropriate for GPS data as it calculates the distance between 2 points on a sphere.
  3. Dispersion of movement: this metric quantifies the extent of the area a patient visited each day. We assessed the spread of GPS coordinates by measuring longitudinal and latitudinal variance and then taking the root mean square sum of these variances.

These metrics were chosen to represent different dimensions of mobility: frequency of travel, distance traveled, and area covered. Together, they provide a comprehensive view of a patient’s mobility patterns and are well-suited to explore the relationship between mobility and COVID-19 risk.

To account for variations among users and days, the raw GPS data were reorganized into daily time series for each participant. Both daily travel distance and daily travel dispersion were then quantile-normalized. Further details, including data preprocessing steps, can be found in Multimedia Appendix 1 (location data preprocessing). COVID-19 epidemiological data COVID-19 epidemiological data encompass global-, national-, and regional-level information, offering insights into various variables such as daily confirmed cases, vaccination rates, and critical patient ratios. For global- and national-level data, we sourced information from Our World in Data, which aggregated data from esteemed institutions such as Johns Hopkins University and the World Health Organization. These data were used to formulate derived variables, including metrics such as the incidence of new cases, vaccination rates, and hospitalizations over a 6-month period, and capture the plausible emergence of herd immunity. Regional-level data, specifically at the province/city level, were obtained from data announced by the Korean health authorities. Data within the same category have undergone minimum-maximum normalization.

Models and Training

Model Selection

Overall, 5 models widely used in the medical field, particularly in relevant studies, have been selected [29-31]. The aim is to showcase that our approach can enhance the performance of various models, regardless of their architectural nuances or inherent biases, as each model inherently carries its distinctive inductive bias. We chose 1 statistical method, 2 classical ML methods, and 2 deep learning methods. The following provides concise explanations of the selected models:

  1. Logistic regression [32,33]: a statistical model that combines the weighted sum of variables with a sigmoid function, providing a straightforward understanding of the relationship between variables and outcomes.
  2. Light gradient boosting machine [34] and extreme gradient boosting [35]: classical ML models widely used for structured table-form data. They leverage powerful algorithms for boosting decision trees, enabling accurate predictions.
  3. Tabular data network (TabNet) [36]: a deep-learning model specifically designed for tabular data. TabNet uses a unique architecture called “attentive transformer” and uses unsupervised learning to enhance performance.
  4. Google AutoML [37]: we incorporated Google’s AutoML, using the functionalities provided by Google Cloud’s Vertex AI. AutoML offers the unique capability to autonomously generate customized neural network architectures. The AutoML model was trained using the same training, validation, and test datasets as the other table-form learning models (Multimedia Appendix 1).
Training Setup

We collected a comprehensive dataset consisting of episodes related to COVID-19 cases, including both positive and negative episodes. To address any potential disparity in the number of positive and negative episodes, we combined the dataset and divided it for training and testing in an 8:2 ratio. To minimize confounding factors, episodes from the same individual were grouped.

To ensure a fair comparison and effective model training, we kept the test set fixed and randomly sampled the validation and training sets. For the models that did not require a validation set, 80% of the data were randomly selected from the training set. The entire experiment was repeated 5 times to account for inherent variability, and the averages and standard deviations of the results were calculated.

Some models used in this study do not inherently support time-series data, so we adopted a modified approach for data extraction. For each episode, we extracted data from the day nearest to the COVID-19 test date, creating a single observation per episode. This approach allowed us to capture relevant information within a limited timeframe while maintaining a consistent evaluation framework. For nontime-series models (logistic regression, light gradient boosting machine, and extreme gradient boosting), this single-day extraction was necessary, while time-series models (TabNet and Google AutoML) could naturally handle sequential data but were evaluated using the same extracted data points for fair comparison (Multimedia Appendix 1).

Performance Evaluation

To consistently evaluate the models’ performance, we used a comprehensive set of metrics, including the area under the receiver operating characteristic curve, area under the precision-recall curve, and F1-score.

Model Interpretability With Shapley Additive Explanations

We used Shapley additive explanations (SHAP) [38] to interpret model predictions. After training and evaluating all 5 algorithms, we selected the best-performing model—Google AutoML trained with the full (“All”) feature set—and calculated mean SHAP values for the entire test set. This provided global explanations of the model’s behavior, showing how symptoms, mobility, and epidemiological features contributed to predictions.

SHINE Dataset Characteristics

A total of 48,798 unique users reported 3,571,243 daily self-checked symptoms and provided basic information, such as age and sex in the SHINE dataset, from October 2020 to March 2023. Over the same period, 17,298 unique users reported 21,773 authorized COVID-19 test results. A total of 15,351 patients reported at least 1 self-reported symptom and an authorized COVID-19 test result, irrespective of their test results. Overall, 5119 negative episodes (33.35%) and 10,232 positive episodes (66.65%) were recorded. Data characteristics are summarized in Table 1.

Table 1. Characteristics of the Study of Health Information for Next Epidemic dataset episodes used for model development.
CategorySHINEa dataset (N=15,376)
Age (years), n (%)
<6014,427 (93.98)
≥60924 (6.02)
Sex, n (%)
Female9928 (64.56)
Male5423 (35.26)
Test indication, n (%)b
Others11,259 (73.34)
Contactc3985 (25.96)
Abroadd107 (0.70)
Test result, n (%)
Negative5119 (33.35)
Positive10,232 (66.65)

aSHINE: Study of Health Information for Next Epidemic.

bReasons prompting COVID-19 testing.

cClose contact with a confirmed COVID-19 case.

dRecent return from international travel.

The data covered a period of 2 weeks before and after the authorized COVID-19 test date, with most episodes lasting within 5 days. The SHINE app was predominantly used by the users to monitor their conditions after the test. We observed a sex imbalance in the dataset, with a significantly higher proportion of females than males (9928 female patients [64.56%] vs 5423 male patients [35.26%]). Comparison of the COVID-19 confirmation rate between sexes revealed no significant differences.

Symptoms in the SHINE Dataset

The symptom distribution was not significantly different by age. All symptoms were significantly more prevalent in the positive group, suggesting robust differences between positive and negative episodes, as shown in Table 2.

Table 2. Comparison of symptom prevalence between the COVID-19–positive and -negative groups in the Study of Health Information for Next Epidemic Dataset. The Mann-Whitney test results indicate a significantly higher prevalence of all symptoms (P<.001) in the positive group.
Symptoms, n (%)Positive group (n=13,977)Negative group (n=6608)PPVa
Cough13,542 (96.88)3652 (55.29)0.7770
Sore throat11,529 (82.47)2161 (32.69)0.8373
Headache10,788 (77.21)1986 (30.04)0.8394
Sputum8290 (59.29)906 (13.72)0.9033
Runny nose4960 (35.51)823 (12.46)0.8551
Chills1908 (13.66)174 (2.63)0.9172
Muscle pain1872 (13.39)302 (4.57)0.8639
Fever3913 (28.01)302 (4.57)0.1234
Shortness of breath951 (6.81)95 (1.44)0.9034
Loss of taste960 (6.87)55 (0.83)0.9392
Loss of smell1000 (7.15)58 (0.88)0.9384
Diarrhea930 (6.59)107 (1.57)0.8968

aPPV: positive predictive value.

In addition, the symptoms were correlated, as shown in Figure 2. For instance, there was a relatively strong correlation between loss of smell and loss of taste (Spearman correlation coefficient: 0.6) and a moderately positive correlation between chills and fever (Spearman correlation coefficient: 0.33). In addition, there were relatively weak correlations among upper respiratory tract symptoms, including cough, sputum production, and sore throat.

In Figure 3, a comparative analysis of the average confirmation rates in Korea over a span of 7 days showed a robust correlation between the asymptomatic confirmation rate and the overall confirmation rate (Spearman rank correlation coefficient=0.7141; P<.001). The asymptomatic confirmation rate was defined as the ratio of users reporting COVID-19 infections without exhibiting any symptoms to the total number of users who reported COVID-19 infections, irrespective of their symptom profiles.

Figure 2. Symptom correlation in the Study of Health Information for Next Epidemic dataset. The Spearman correlation coefficients are displayed in the grid. There was a relatively strong correlation between loss of smell and loss of taste (0.6) and a moderately positive correlation between chills and fever (0.33). While some symptoms tend to co-occur more frequently, the overall pattern of symptom presentation is varied.
Figure 3. Comparison of 7-day moving average values for confirmed cases and asymptomatic positive ratio in the Study of Health Information for Next Epidemic dataset. The Spearman rank correlation coefficient between the 2 metrics is 0.7141 (P<.001), highlighting the persistent presence of asymptomatic cases throughout the pandemic period.

Impact of Mobility and Epidemiological Feature Addition on COVID-19 Prediction in the SHINE Dataset

Using the SHINE dataset, we conducted an experiment aimed to evaluate the improvement in performance achieved by selectively incorporating different groups of features into the baseline model, as shown in Table 3.

Table 3. Model performance in predicting COVID-19 with the integration of secondary features using the Study of Health Information for Next Epidemic dataset. Each row represents model performance after adding the indicated feature groups on top of the baseline.
Metrics (mean)LRaXGBoostbLGBMcTabNetdAutoMLe
Baselinef
AUROCg0.86460.86410.86430.86290.8712
AUPRCh0.92860.92830.92860.92600.8686
F1-score0.85660.85380.85110.85550.8051
+Mobilityi
AUROC0.87320.87510.87970.86920.8926
AUPRC0.93200.93110.93620.92930.8892
F1-score0.86020.85650.85780.85310.8260
+Epidemiologicalj
AUROC0.90050.89780.90110.89560.9060
AUPRC0.94430.94190.94500.94010.9040
F1-score0.87610.87650.87660.69940.8349
Allk
AUROC0.90080.89830.90460.89340.9104
AUPRC0.94360.94130.94710.93780.9084
F1-score0.87850.87450.87690.87300.8355

aLR: logistic regression.

bXGBoost: extreme gradient boosting.

cLGBM: light gradient boosting machine.

dTabNet: tabular network.

eAutoML: automated machine learning.

fBaseline: symptom + demographic variables only.

gAUROC: area under the receiver operating characteristic curve.

hAUPRC: area under the precision-recall curve.

i+Mobility: baseline plus 3 GPS-derived mobility metrics.

j+Epidemiological: baseline plus global·national·regional COVID-19 indices.

kAll: baseline plus both mobility and epidemiological features.

Significant performance improvements were observed when COVID-19 epidemiological data were incorporated into the baseline model, with these improvements surpassing those achieved by integrating the mobility pattern data (mean difference in area under the receiver operating characteristic curve: 0.03 vs 0.01 for all models, Mann-Whitney test; P<.001). Notably, the improvement in model performance was greater with the inclusion of COVID-19 epidemiological data than with the inclusion of mobility pattern data.

SHAP Analysis

In Figure 4, cough emerged as the most influential predictor of COVID-19 positivity with a mean SHAP value of +0.06, followed by national confirmed cases (+0.05) and sputum production (+0.05). Sore throat ranked fourth (+0.04) and global confirmed cases fifth (+0.03). Mobility-related daily travel dispersion (+0.02) and national new deaths (+0.02) also contributed, whereas demographic variables such as age and fever showed only minor effects (+0.01 each). Respiratory symptoms, therefore, headed the importance ranking, with mobility and epidemiological indicators providing meaningful, but comparatively weaker, additional signals.

Figure 4. Top 10 important features, ranked by mean Shapley additive explanations values (Google AutoML model, “All” feature set) in the Study of Health Information for Next Epidemic dataset. National confirmed cases, global confirmed cases, and national new deaths are normalized values per million people. Daily travel dispersion refers to the travel patterns on the day prior to the test date. SHAP: Shapley additive explanations.

Principal Findings

In this study, the accuracy of ML models for COVID-19 screening was significantly improved by the incorporation of external data, namely, mobility and epidemiological data, from the SHINE datasets. Our preliminary work with the Israeli dataset (Multimedia Appendix 1) provided initial support for this finding. The attribution of features in enhancing model performance was further reinforced by our SHAP analysis, which provided explainable evidence for the value of contextual data.

SHAP analysis showed that respiratory symptoms, such as cough, exerted the strongest influence on model predictions, whereas epidemiological and mobility variables ranked immediately after these symptoms and provided complementary but comparatively weaker predictive value. This highlights a key limitation of symptom-based screening: while symptoms such as fever or cough were significantly more prevalent in patients who tested positive for COVID-19, their overlap with other illnesses and the existence of thousands of asymptomatic positive cases reduced their diagnostic specificity. The synergy between contextual data (eg, exposure history, population movement patterns) and clinical features demonstrates that integrating diverse data sources—rather than relying on symptoms alone—can substantially improve screening accuracy and model reliability, particularly in asymptomatic or presymptomatic populations.

Limitations

Our findings have certain limitations from the perspective of patient-generated health data. Although the volume of self-checked data collected was consistently maintained throughout the study period, we observed a concentrated accumulation of COVID-19 test submissions during the 2-month period between March 2022 and April 2022, in contrast to the even distribution of self-checked symptom submissions throughout the study duration. The higher number of tests recorded during this concentrated period can be attributed to increased awareness and concern regarding the use of the app and testing in the population. This is not particularly unusual, but it is important to consider that it may have contributed to model learning from our data during this specific period. Further, our data may have a selection bias due to government policies. For instance, during the data collection period, it was not possible to board return flights if individuals were infected with COVID-19 while overseas [20,39]. This suggests that in the future, when developing similar models for outbreaks of unprecedented epidemics such as the COVID-19 pandemic, incorporating the indicator of being a returning traveler from overseas may lead to a potential misjudgment of positivity probability.

According to our findings, incorporating mobility data has a positive impact on the performance of ML models. However, the metrics are subject to retrospective interpretation [40-42]. For instance, individuals who have already displayed severe initial symptoms of COVID-19 may find it more challenging to move actively. Furthermore, in our methodology, we focused on quantifying the extent of individuals’ activity levels rather than assessing the number of high-risk areas visited within that activity range.

Conclusions

In conclusion, our study demonstrated the transformative potential of integrating mobility and epidemiological data into ML models for COVID-19 prediction. These external data sources not only improved the model accuracy but also highlighted the limitations of symptom-based predictions and the value of comprehensive data analysis. Our insights have practical implications for health care decision-making and public health interventions, paving the way for more informed responses to outbreaks.

Acknowledgments

HJK contributed equally to the supervision of this work alongside MK.

We thank Ahreum Jang, Hyejung Kim, Hae-Lee Park, and Sungtae Kim for their contributions to this work.

Funding

This study is based on the research “A Next Generation Surveillance Study for Epidemic Preparedness” which was funded by the Bill & Melinda Gates Foundation (grant number: INV-006404). The funders had no role in the study design, data collection, data analysis, data interpretation, or writing of the report. The findings and conclusions contained within are those of the authors and do not necessarily reflect positions or policies of the Bill & Melinda Gates Foundation.

Data Availability

The Study of Health Information for Next Epidemic (SHINE) research consortium established a SHINE data repository containing datasets generated from our studies. All data in this repository were deidentified. The datasets generated or analyzed during this study are available from the corresponding author on reasonable request. Interested parties can access these datasets after signing a data-sharing agreement and obtaining approval from the consortium, thereby ensuring the use of the data for further research and discovery only. However, owing to certain regulations, the data collected for this study may not be available to external parties. To request access to the data, please visit the SHINE website [43]. The Israeli dataset, which included information on all individuals tested for SARS-CoV-2 via reverse transcription polymerase chain reaction of nasopharyngeal swabs, was publicly released by the Israeli Ministry of Health. The dataset can be accessed via the Israeli Government database Datagov [26]. Data collected until June 30, 2023, were used for this study. Our World in Data, COVID-19 data are publicly available at GitHub Our World in Data web page [45]. We used these data until June 30, 2023, for this study. The code is available at GitHub mobiledoctorDev web page [46].

Authors' Contributions

EK and JL collected and verified the data. JL, EK, and CK designed the SHINE app to acquire patient data. JL, CK, NO, and HJK managed the research. NO and SYS secured funding and other resources for this study. HC, Dohyung Lee, and MK drafted the article. SYS, MK, and HJK revised the article. Duhun L handled the front-end and back-end software development of the app, while Dohyung L and MK developed software for the data pipeline. Dohyung L conducted the preliminary experiments, whereas MK and HC performed the main experiments and statistical analyses. HC created the figures in the article. HJK, MK, and SYS supervised the research, investigated the experimental performance, and collected the data.

Conflicts of Interest

MK, Dohyung L, Duhun L, NO, JL, and EK are employed by Mobile Doctor, with MK and NO holding shares in the company. CK was employed by Mobile Doctor. Mobile Doctor and KT are member entities of the SHINE research consortium. SYS is an employee of Kakao Healthcare.

Multimedia Appendix 1

Information about all variable lists, variable processing method, model hyperparameters, and preliminary experiment on Israeli dataset.

DOCX File, 3928 KB

Checklist 1

CONSORT-AI checklist.

PDF File, 74 KB

  1. Testing for COVID-19. Centers for Disease Control and Prevention (CDC) COVID-19. 2025. URL: https:/​/www.​cdc.gov/​covid/​testing/​?CDC_AAref_Val=https:/​/www.​cdc.gov/​coronavirus/​2019-ncov/​symptoms-testing/​testing.​html [Accessed 2023-10-31]
  2. Esbin MN, Whitney ON, Chong S, Maurer A, Darzacq X, Tjian R. Overcoming the bottleneck to widespread testing: a rapid review of nucleic acid testing approaches for COVID-19 detection. RNA. Jul 2020;26(7):771-783. [CrossRef] [Medline]
  3. Mercer TR, Salit M. Testing at scale during the COVID-19 pandemic. Nat Rev Genet. Jul 2021;22(7):415-426. [CrossRef] [Medline]
  4. Xu Z, Shi L, Wang Y, et al. Pathological findings of COVID-19 associated with acute respiratory distress syndrome. Lancet Respir Med. Apr 2020;8(4):420-422. [CrossRef] [Medline]
  5. Myers JF, Snyder RE, Porse CC, et al. Identification and monitoring of international travelers during the initial phase of an outbreak of COVID-19 - California, February 3-March 17, 2020. MMWR Morb Mortal Wkly Rep. May 15, 2020;69(19):599-602. [CrossRef] [Medline]
  6. Schuchat A, CDC COVID-19 Response Team. Public health response to the initiation and spread of pandemic COVID-19 in the United States, February 24-April 21, 2020. MMWR Morb Mortal Wkly Rep. May 8, 2020;69(18):551-556. [CrossRef] [Medline]
  7. Demicheli V, Jefferson T, Rivetti D, Deeks J. Prevention and early treatment of influenza in healthy adults. Vaccine (Auckl). Jan 6, 2000;18(11-12):957-1030. [CrossRef] [Medline]
  8. Guide for considering influenza testing when influenza viruses are circulating in the community. Centers for Disease Control and Prevention. URL: https:/​/www.​cdc.gov/​flu/​hcp/​testing-methods/​consider-influenza-testing.html?CDC_AAref_Val=https:/​/www.​cdc.gov/​flu/​professionals/​diagnosis/​consider-influenza-testing.​htm [Accessed 2025-10-13]
  9. Halani S, Tombindo PE, O’Reilly R, et al. Clinical manifestations and health outcomes associated with Zika virus infections in adults: a systematic review. PLOS Negl Trop Dis. Jul 2021;15(7):e0009516. [CrossRef] [Medline]
  10. Burger-Calderon R, Bustos Carrillo F, Gresh L, et al. Age-dependent manifestations and case definitions of paediatric Zika: a prospective cohort study. Lancet Infect Dis. Mar 2020;20(3):371-380. [CrossRef] [Medline]
  11. Chandramohan D, Jaffar S, Greenwood B. Use of clinical algorithms for diagnosing malaria. Trop Med Int Health. Jan 2002;7(1):45-52. [CrossRef] [Medline]
  12. Levine AC, Shetty PP, Burbach R, et al. Derivation and internal validation of the ebola prediction score for risk stratification of patients with suspected ebola virus disease. Ann Emerg Med. Sep 2015;66(3):285-293. [CrossRef] [Medline]
  13. Zoabi Y, Deri-Rozov S, Shomron N. Machine learning-based prediction of COVID-19 diagnosis based on symptoms. NPJ Digit Med. Jan 4, 2021;4(1):3. [CrossRef] [Medline]
  14. Kennedy B, Fitipaldi H, Hammar U, et al. App-based COVID-19 syndromic surveillance and prediction of hospital admissions in COVID symptom study Sweden. Nat Commun. Apr 21, 2022;13(1):2110. [CrossRef] [Medline]
  15. Gostic K, Gomez AC, Mummah RO, Kucharski AJ, Lloyd-Smith JO. Estimated effectiveness of symptom and risk screening to prevent the spread of COVID-19. Elife. Feb 24, 2020;9:e55570. [CrossRef] [Medline]
  16. Lai CC, Shih TP, Ko WC, Tang HJ, Hsueh PR. Severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) and coronavirus disease-2019 (COVID-19): the epidemic and the challenges. Int J Antimicrob Agents. Mar 2020;55(3):105924. [CrossRef] [Medline]
  17. Solomon DA, Sherman AC, Kanjilal S. Influenza in the COVID-19 era. JAMA. Oct 6, 2020;324(13):1342-1343. [CrossRef] [Medline]
  18. Callahan A, Steinberg E, Fries JA, et al. Estimating the efficacy of symptom-based screening for COVID-19. NPJ Digit Med. 2020;3(1):95. [CrossRef] [Medline]
  19. Chow EJ, Schwartz NG, Tobolowsky FA, et al. Symptom screening at illness onset of health care personnel with SARS-CoV-2 infection in King County, Washington. JAMA. May 26, 2020;323(20):2087-2089. [CrossRef] [Medline]
  20. Bae SH, Shin H, Koo HY, Lee SW, Yang JM, Yon DK. Asymptomatic transmission of SARS-CoV-2 on evacuation flight. Emerg Infect Dis. Nov 2020;26(11):2705-2708. [CrossRef] [Medline]
  21. Moran KR, Fairchild G, Generous N, et al. Epidemic forecasting is messier than weather forecasting: the role of human behavior and internet data streams in epidemic forecast. J Infect Dis. Dec 1, 2016;214(suppl_4):S404-S408. [CrossRef] [Medline]
  22. Eubank S, Guclu H, Kumar VSA, et al. Modelling disease outbreaks in realistic urban social networks. Nature New Biol. May 13, 2004;429(6988):180-184. [CrossRef] [Medline]
  23. Longini IM Jr, Nizam A, Xu S, et al. Containing pandemic influenza at the source. Science. Aug 12, 2005;309(5737):1083-1087. [CrossRef] [Medline]
  24. Choo H, Kim M, Choi J, Shin J, Shin SY. Influenza screening via deep learning using a combination of epidemiological and patient-generated health data: development and validation study. J Med Internet Res. Oct 29, 2020;22(10):e21369. [CrossRef] [Medline]
  25. Loo BPY, Tsoi KH, Wong PPY, Lai PC. Identification of superspreading environment under COVID-19 through human mobility data. Sci Rep. Feb 25, 2021;11(1):4699. [CrossRef] [Medline]
  26. COVID-19 database (Web page in Hebrew). Data Gov Government Databases. 2022. URL: https://data.gov.il/dataset/covid-19 [Accessed 2023-10-31]
  27. SHINE (study of health information for next epidemic). SHINE. URL: https://shineforall.org/eng/ [Accessed 2023-10-31]
  28. COVID-19 Pandemic. Our World in Data. URL: https://ourworldindata.org/coronavirus [Accessed 2023-10-31]
  29. Chahar S, Roy PK. COVID-19: a comprehensive review of learning models. Arch Comput Methods Eng. 2022;29(3):1915-1940. [CrossRef] [Medline]
  30. Ikemura K, Bellin E, Yagi Y, et al. Using automated machine learning to predict the mortality of patients with COVID-19: prediction model development study. J Med Internet Res. Feb 26, 2021;23(2):e23458. [CrossRef] [Medline]
  31. Nazir A, Ampadu HK. Interpretable deep learning for the prediction of ICU admission likelihood and mortality of COVID-19 patients. PeerJ Comput Sci. 2022;8:e889. [CrossRef] [Medline]
  32. Cox DR. The regression analysis of binary sequences. J R Stat Soc Ser B Stat Methodol. Jan 1, 1959;21(1):238-238. [CrossRef]
  33. Varoquaux G, Buitinck L, Louppe G, Grisel O, Pedregosa F, Mueller A. Scikit-learn: machine learning without learning the machinery. GetMobile Mob Comput Commun. 2015;19(1):29-33. [CrossRef]
  34. Ke G, Meng Q, Finley T, et al. LightGBM: a highly efficient gradient boosting decision tree. Presented at: NIPS’17: Proceedings of the 31st International Conference on Neural Information Processing Systems; Dec 4-9, 2017. [CrossRef]
  35. Chen T, Guestrin C. XGBoost: a scalable tree boosting system. Presented at: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining - KDD ’16; Aug 13-17, 2016:785-794; San Francisco, CA. [CrossRef]
  36. Arik S, Pfister T. TabNet: attentive interpretable tabular learning. Presented at: Proceedings of the AAAI Conference on Artificial Intelligence; Feb 2-9, 2021:6679-6687; Sunnyvale, CA. [CrossRef]
  37. An end-to-end automl solution for tabular data at kaggledays. Google AI Blog. 2019. URL: https://ai.googleblog.com/2019/05/an-end-to-end-automl-solution-for.html [Accessed 2023-10-31]
  38. Lundberg S, Lee SI. A unified approach to interpreting model predictions. Presented at: 31st Conference on Neural Information Processing Systems (NIPS 2017); Dec 4-9, 2017. [CrossRef]
  39. Pre-departure negative COVID-19 test result requirement for international arrival (12.30, 2021 updated). Consulate General of the Republic of Korea in Atlanta. 2021. URL: https://overseas.mofa.go.kr/us-atlanta-en/brd/m_4842/view.do?seq=761311 [Accessed 2023-09-02]
  40. Kishore N, Taylor AR, Jacob PE, et al. Evaluating the reliability of mobility metrics from aggregated mobile phone data as proxies for SARS-CoV-2 transmission in the USA: a population-based study. Lancet Digit Health. Jan 2022;4(1):e27-e36. [CrossRef] [Medline]
  41. Jewell S, Futoma J, Hannah L, Miller AC, Foti NJ, Fox EB. It’s complicated: characterizing the time-varying relationship between cell phone mobility and COVID-19 spread in the US. NPJ Digit Med. Oct 27, 2021;4(1):152. [CrossRef] [Medline]
  42. Badr HS, Du H, Marshall M, Dong E, Squire MM, Gardner LM. Association between mobility patterns and COVID-19 transmission in the USA: a mathematical modelling study. Lancet Infect Dis. Nov 2020;20(11):1247-1254. [CrossRef] [Medline]
  43. Data. SHINE. URL: https://shineforall.org/eng/analysis/ [Accessed 2026-01-06]
  44. Our world in data. GitHub. URL: https://github.com/owid/covid-19-data [Accessed 2026-01-06]
  45. MobiledoctorDev. GitHub. URL: https://github.com/mobiledoctorDev/SHINE_patient_is_not_all_you_need [Accessed 2026-01-06]


CONSORT-AI: Consolidated Standards of Reporting Trials–Artificial Intelligence
ML: machine learning
PCR: polymerase chain reaction
SHAP: Shapley additive explanations
SHINE: Study of Health Information for Next Epidemic
TabNet: tabular data network


Edited by Khaled El Emam; submitted 11.Dec.2023; peer-reviewed by Yaoping Ruan, Yizhen Li; final revised version received 02.Aug.2025; accepted 08.Aug.2025; published 05.Mar.2026.

Copyright

© Hyunwoo Choo, Dohyung Lee, Soo-Yong Shin, Jiwoo Lee, Duhun Lee, Eonji Kim, Namsoo Oh, Christina Kim, Myeongchan Kim, Hyo Jung Kim. Originally published in JMIR AI (https://ai.jmir.org), 5.Mar.2026.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR AI, is properly cited. The complete bibliographic information, a link to the original publication on https://www.ai.jmir.org/, as well as this copyright and license information must be included.